Goto

Collaborating Authors

 base heuristic


Collapse and Collision Aware Grasping for Cluttered Shelf Picking

Pathak, Abhinav, Muthusamy, Rajkumar

arXiv.org Artificial Intelligence

-- Efficient and safe retrieval of stacked objects in warehouse environments is a significant challenge due to complex spatial dependencies and structural inter-dependencies. Traditional vision-based methods excel at object localization but often lack the physical reasoning required to predict the consequences of extraction, leading to unintended collisions and collapses. This paper proposes a collapse and collision aware grasp planner that integrates dynamic physics simulations for robotic decision-making. Using a single image and depth map, an approximate 3D representation of the scene is reconstructed in a simulation environment, enabling the robot to evaluate different retrieval strategies before execution. Two approaches 1) heuristic-based and 2) physics-based are proposed for both single-box extraction and shelf clearance tasks. Extensive real-world experiments on structured and unstructured box stacks, along with validation using datasets from existing databases, show that our physics-aware method significantly improves efficiency and success rates compared to baseline heuristics. The integration of autonomous robotic systems into warehouse management has led to significant improvements in efficiency, particularly in tasks such as item retrieval, transportation, and inventory organization.


An Approximate Dynamic Programming Framework for Occlusion-Robust Multi-Object Tracking

Musunuru, Pratyusha, Li, Yuchao, Weber, Jamison, Bertsekas, Dimitri

arXiv.org Artificial Intelligence

In this work, we consider data association problems involving multi-object tracking (MOT). In particular, we address the challenges arising from object occlusions. We propose a framework called approximate dynamic programming track (ADPTrack), which applies dynamic programming principles to improve an existing method called the base heuristic. Given a set of tracks and the next target frame, the base heuristic extends the tracks by matching them to the objects of this target frame directly. In contrast, ADPTrack first processes a few subsequent frames and applies the base heuristic starting from the next target frame to obtain tentative tracks. It then leverages the tentative tracks to match the objects of the target frame. This tends to reduce the occlusion-based errors and leads to an improvement over the base heuristic. When tested on the MOT17 video dataset, the proposed method demonstrates a 0.7% improvement in the association accuracy (IDF1 metric) over a state-of-the-art method that is used as the base heuristic. It also obtains improvements with respect to all the other standard metrics. Empirically, we found that the improvements are particularly pronounced in scenarios where the video data is obtained by fixed-position cameras.


Longest Common Substring in Longest Common Subsequence's Solution Service: A Novel Hyper-Heuristic

Abdi, Alireza, Hajsaeedi, Masih, Hooshmand, Mohsen

arXiv.org Artificial Intelligence

The Longest Common Subsequence (LCS) is the problem of finding a subsequence among a set of strings that has two properties of being common to all and is the longest. The LCS has applications in computational biology and text editing, among many others. Due to the NP-hardness of the general longest common subsequence, numerous heuristic algorithms and solvers have been proposed to give the best possible solution for different sets of strings. None of them has the best performance for all types of sets. In addition, there is no method to specify the type of a given set of strings. Besides that, the available hyper-heuristic is not efficient and fast enough to solve this problem in real-world applications. This paper proposes a novel hyper-heuristic to solve the longest common subsequence problem using a novel criterion to classify a set of strings based on their similarity. To do this, we offer a general stochastic framework to identify the type of a given set of strings. Following that, we introduce the set similarity dichotomizer ($S^2D$) algorithm based on the framework that divides the type of sets into two. This algorithm is introduced for the first time in this paper and opens a new way to go beyond the current LCS solvers. Then, we present a novel hyper-heuristic that exploits the $S^2D$ and one of the internal properties of the set to choose the best matching heuristic among a set of heuristics. We compare the results on benchmark datasets with the best heuristics and hyper-heuristics. The results show a higher performance of our proposed hyper-heuristic in both quality of solutions and run time factors.


Reinforcement Learning Methods for Wordle: A POMDP/Adaptive Control Approach

Bhambri, Siddhant, Bhattacharjee, Amrita, Bertsekas, Dimitri

arXiv.org Artificial Intelligence

In this paper, we discuss a Reinforcement Learning (RL) approach towards a class of sequential decision problems, exemplified for the popular Wordle puzzle that appears daily in the New York Times. Wordle involves a list of 5-letter mystery words, which is a subset of a larger list of guess words. A word is selected at random from the mystery list, and the objective is to find that word by sequentially selecting no more than six words from the guess list. Each guess word selection provides information about the letters contained in the hidden mystery word according to a given set of rules, which involves color coding of letters shared by the guess word and the mystery word. We will adopt a more general point of view, by considering a broad class of problems that include Wordle as a special case. In particular, the problems that we consider include sequential search situations, where the objective is to guess correctly an unknown object from a given finite set of objects (the set of mystery words in the Wordle context), by using a sequence of decisions from a finite set (the set of guess words in Wordle), which result in a sequence of corresponding observations (the information outcomes of the guesses in Wordle). We aim to minimize some cost function, such as the expected number of observations required to determine the unknown object. Within the search context just described, some basic information theory concepts are relevant, which have already been applied to Wordle, and are important for our methodology.


Constrained Multiagent Rollout and Multidimensional Assignment with the Auction Algorithm

Bertsekas, Dimitri

arXiv.org Artificial Intelligence

We consider an extension of the rollout algorithm that applies to constrained deterministic dynamic programming, including challenging combinatorial optimization problems. The algorithm relies on a suboptimal policy, called base heuristic. Under suitable assumptions, we show that if the base heuristic produces a feasible solution, the rollout algorithm has a cost improvement property: it produces a feasible solution, whose cost is no worse than the base heuristic's cost. We then focus on multiagent problems, where the control at each stage consists of multiple components (one per agent), which are coupled either through the cost function or the constraints or both. We show that the cost improvement property is maintained with an alternative implementation that has greatly reduced computational requirements, and makes possible the use of rollout in problems with many agents. We demonstrate this alternative algorithm by applying it to layered graph problems that involve both a spatial and a temporal structure. We consider in some detail a prominent example of such problems: multidimensional assignment, where we use the auction algorithm for 2-dimensional assignment as a base heuristic. This auction algorithm is particularly well-suited for our context, because through the use of prices, it can advantageously use the solution of an assignment problem as a starting point for solving other related assignment problems, and this can greatly speed up the execution of the rollout algorithm.


The Compressed Differential Heuristic

Goldenberg, Meir (Ben-Gurion University) | Sturtevant, Nathan (University of Denver) | Felner, Ariel (Ben-Gurion University) | Schaeffer, Jonathan (University of Alberta)

AAAI Conferences

The differential heuristic (DH) is an effective memory-based heuristic for explicit state spaces. In this paper we aim to improve its performance and memory usage. We introduce a compression method for DHs which stores only a portion of the original uncompressed DH, while preserving enough information to enable efficient search. Compressed DHs (CDH) are flexible and can be tuned to fit any size of memory, even smaller than the size of the state space. Furthermore, CDHs can be built without the need to create and store the entire uncompressed DH. Experimental results across different domains show that, for a given amount of memory, a CDH significantly outperforms an uncompressed DH.


Learning Inadmissible Heuristics During Search

Thayer, Jordan Tyler (University of New Hampshire) | Dionne, Austin (University of New Hampshire) | Ruml, Wheeler (University of New Hampshire)

AAAI Conferences

Suboptimal search algorithms offer shorter solving times by sacrificing guaranteed solution optimality. While optimal searchalgorithms like A* and IDA* require admissible heuristics, suboptimalsearch algorithms need not constrain their guidance in this way. Previous work has explored using off-line training to transform admissible heuristics into more effective inadmissible ones. In this paper we demonstrate that this transformation can be performed on-line, during search. In addition to not requiring training instances and extensive pre-computation, an on-line approach allows the learned heuristic to be tailored to a specific problem instance. We evaluate our techniques in four different benchmark domains using both greedy best-first search and bounded suboptimal search. We find that heuristics learned on-line result in both faster search andbetter solutions while relying only on information readily available in any best-first search.